Partially Observable Minimum-Age Scheduling: The Greedy Policy

نویسندگان

چکیده

This paper studies the minimum-age scheduling problem in a wireless sensor network where an access point (AP) monitors state of object via set sensors. The freshness sensed state, measured by age-of-information (AoI), varies at different sensors and is not directly observable to AP. AP has decide which query/sample order get most updated information (i.e., with minimum AoI). In this paper, we formulate as multi-armed bandit partially arms explore greedy policy minimize expected AoI sampled over infinite horizon. To analyze performance policy, 1) put forth relaxed that decouples sampling processes arms, 2) process each arm Markov decision (POMDP), 3) derive average under sum from individual arms. Numerical simulation results validate excellent approximation terms

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Correctness of the greedy algorithm for minimum lateness scheduling

We are given n jobs 1, 2, . . . n, and a release time r. Each job i has duration t(i) > 0 and deadline d(i). A schedule s defines, for each job i, a start time s(i) ≥ r so that for any two distinct jobs i and j the intervals [s(i), s[i] + t(i)] and [s(j), s(j) + t(j)] do not overlap (one finishes no later than the other starts). The lateness l(i) of job i in schedule s is the amount of time by ...

متن کامل

Policy Filtering for Planning in Partially Observable Stochastic Domains

Partially observable Markov decision processes (POMDP) can be used as a model for planning in stochastic domains. This paper considers the problem of computing optimal policies for nite horizon POMDPs. In deciding on an action to take, an agent is not only concerned with how the action would a ect the current time point, but also its impacts on the rest of the planning horizon. In a POMDP, the ...

متن کامل

Policy-Gradient Algorithms for Partially Observable Markov Decision Processes

Partially observable Markov decision processes are interesting because of their ability to model most conceivable real-world learning problems, for example, robot navigation, driving a car, speech recognition, stock trading, and playing games. The downside of this generality is that exact algorithms are computationally intractable. Such computational complexity motivates approximate approaches....

متن کامل

An Improved Policy Iteration Algorithm for Partially Observable MDPs

A new policy iteration algorithm for partially observable Markov decision processes is presented that is simpler and more eecient than an earlier policy iteration algorithm of Sondik (1971,1978). The key simpliication is representation of a policy as a nite-state controller. This representation makes policy evaluation straightforward. The pa-per's contribution is to show that the dynamic-progra...

متن کامل

A Greedy Approximation Algorithm for Minimum-Gap Scheduling

We consider scheduling of unit-length jobs with release times and deadlines, where the objective is to minimize the number of gaps in the schedule. Polynomial-time algorithms for this problem are known, yet they are rather inefficient, with the best algorithm running in time O(n) and requiring O(n) memory. We present a greedy algorithm that approximates the optimum solution within a factor of 2...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Communications

سال: 2022

ISSN: ['1558-0857', '0090-6778']

DOI: https://doi.org/10.1109/tcomm.2021.3123362